# Multi-bit Quantization

INTELLECT 2 GGUF
INTELLECT-2-GGUF is the GGUF format quantized version of PrimeIntellect/INTELLECT-2, suitable for text generation tasks.
Large Language Model
I
MaziyarPanahi
88
1
Gemma 3 27b It GGUF
GGUF quantized version of the Gemma 3 27B instruction-tuned model, suitable for text generation tasks.
Large Language Model
G
MaziyarPanahi
475.91k
4
Gemma 3 4b It GGUF
GGUF quantized version of the Gemma 3B model, suitable for local text generation tasks
Large Language Model
G
MaziyarPanahi
358.91k
6
Mistral Small 24B Instruct 2501 GGUF
GGUF quantized version of Mistral-Small-24B-Instruct-2501, suitable for local deployment and text generation tasks.
Large Language Model
M
MaziyarPanahi
474.73k
2
Llama 3.2 3B Instruct Abliterated GGUF
GGUF-format quantized version of Llama-3.2-3B-Instruct-abliterated, supporting multiple bit quantization options, suitable for text generation tasks.
Large Language Model
L
MaziyarPanahi
181
2
Yi Coder 1.5B Chat GGUF
Yi-Coder-1.5B-Chat-GGUF is the GGUF format model file of 01-ai/Yi-Coder-1.5B-Chat, suitable for text generation tasks.
Large Language Model
Y
MaziyarPanahi
254.78k
10
Phi 3.5 Mini Instruct GGUF
GGUF format model file for Phi-3.5-mini-instruct, suitable for text generation tasks.
Large Language Model
P
MaziyarPanahi
335.88k
13
Mistral Nemo Instruct 2407 GGUF
Mistral-Nemo-Instruct-2407-GGUF is the GGUF format quantized version of mistralai/Mistral-Nemo-Instruct-2407, supporting multiple quantization bits (2-bit to 8-bit), suitable for text generation tasks.
Large Language Model
M
MaziyarPanahi
441.17k
44
Mistral 7B Instruct V0.3 GGUF
Apache-2.0
GGUF quantized version of Mistral-7B-Instruct-v0.3, suitable for local inference text generation models
Large Language Model
M
MaziyarPanahi
253.99k
93
Zephyr Orpo 141b A35b V0.1 GGUF
Apache-2.0
A 141-billion parameter Mixture of Experts (MoE) model fine-tuned from Mixtral-8x22B-v0.1, with 35 billion active parameters, primarily designed for English text generation tasks
Large Language Model English
Z
MaziyarPanahi
10.04k
29
Saul Instruct V1 GGUF
MIT
Saul-Instruct-v1-GGUF is the GGUF format version of the Equall/Saul-Instruct-v1 model, suitable for text generation tasks and supports multiple quantization levels.
Large Language Model English
S
MaziyarPanahi
456
8
Gemma 2b It GGUF
Other
GGUF quantized version of the Gemma 2B model, suitable for local deployment and inference
Large Language Model
G
MaziyarPanahi
517
10
Rank Zephyr 7b V1 Full GGUF
MIT
A text ranking model based on Mistral-7B, offering multiple quantized versions for efficient inference.
Large Language Model English
R
MaziyarPanahi
708
5
Sauerkrautlm 7b HerO Mistral 7B Instruct V0.1 GGUF
Apache-2.0
This is a German/English bilingual model fine-tuned based on Mistral-7B-Instruct-v0.1, quantized in GGUF format with support for multiple quantization levels from 2 to 8 bits.
Large Language Model Supports Multiple Languages
S
MaziyarPanahi
90
2
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase